Bradley County
A$^2$Search: Ambiguity-Aware Question Answering with Reinforcement Learning
Zhang, Fengji, Niu, Xinyao, Ying, Chengyang, Lin, Guancheng, Hao, Zhongkai, Fan, Zhou, Huang, Chengen, Keung, Jacky, Chen, Bei, Lin, Junyang
Recent advances in Large Language Models (LLMs) and Reinforcement Learning (RL) have led to strong performance in open-domain question answering (QA). However, existing models still struggle with questions that admit multiple valid answers. Standard QA benchmarks, which typically assume a single gold answer, overlook this reality and thus produce inappropriate training signals. Existing attempts to handle ambiguity often rely on costly manual annotation, which is difficult to scale to multi-hop datasets such as HotpotQA and MuSiQue. In this paper, we present A$^2$Search, an annotation-free, end-to-end training framework to recognize and handle ambiguity. At its core is an automated pipeline that detects ambiguous questions and gathers alternative answers via trajectory sampling and evidence verification. The model is then optimized with RL using a carefully designed $\mathrm{AnsF1}$ reward, which naturally accommodates multiple answers. Experiments on eight open-domain QA benchmarks demonstrate that A$^2$Search achieves new state-of-the-art performance. With only a single rollout, A$^2$Search-7B yields an average $\mathrm{AnsF1}@1$ score of $48.4\%$ across four multi-hop benchmarks, outperforming all strong baselines, including the substantially larger ReSearch-32B ($46.2\%$). Extensive analyses further show that A$^2$Search resolves ambiguity and generalizes across benchmarks, highlighting that embracing ambiguity is essential for building more reliable QA systems. Our code, data, and model weights can be found at https://github.com/zfj1998/A2Search
- Europe > Russia (0.14)
- Asia > Russia (0.14)
- North America > United States > New York (0.04)
- (30 more...)
- Research Report > New Finding (1.00)
- Personal > Obituary (1.00)
- Workflow (0.93)
- Media > Music (1.00)
- Media > Film (1.00)
- Leisure & Entertainment (1.00)
- (5 more...)
Aligning Multiple Knowledge Graphs in a Single Pass
Yang, Yaming, Wang, Zhe, Guan, Ziyu, Zhao, Wei, Lu, Weigang, Huang, Xinyan
Entity alignment (EA) is to identify equivalent entities across different knowledge graphs (KGs), which can help fuse these KGs into a more comprehensive one. Previous EA methods mainly focus on aligning a pair of KGs, and to the best of our knowledge, no existing EA method considers aligning multiple (more than two) KGs. To fill this research gap, in this work, we study a novel problem of aligning multiple KGs and propose an effective framework named MultiEA to solve the problem. First, we embed the entities of all the candidate KGs into a common feature space by a shared KG encoder. Then, we explore three alignment strategies to minimize the distances among pre-aligned entities. In particular, we propose an innovative inference enhancement technique to improve the alignment performance by incorporating high-order similarities. Finally, to verify the effectiveness of MultiEA, we construct two new real-world benchmark datasets and conduct extensive experiments on them. The results show that our MultiEA can effectively and efficiently align multiple KGs in a single pass.
- Europe > Austria > Vienna (0.14)
- Asia > China > Shaanxi Province > Xi'an (0.04)
- Europe > Russia (0.04)
- (11 more...)